Speech generating device

Speech generating devices (SGD), also known as voice output communication aids, are electronic augmentative and alternative communication (AAC) systems used to supplement or replace speech or writing for individuals with severe speech impairments, enabling them to verbally communicate their needs.[1] SGDs are important for people who have limited means of interacting verbally, as they allow individuals to become active participants in communication interactions.[2]

A variety of different input and display methods exist for users of varying abilities to make use of SGDs. Some SGDs have multiple pages of symbols to accommodate a large number of utterances, and thus only a portion of the symbols available are visible at any one time, with the communicator navigating the various pages. Speech generating devices can produce electronic voice output using by digitized recording of natural speech, or by speech synthesis, which may carry less emotional information but can permit the user to speak novel messages .[3]

The content, organisation, and updating of this vocabulary on a SGD is influenced by a number of factors, such at the user's needs and the contexts that the device will be used in.[4] The development of techniques to improve the available vocabulary and rate of speech production is an active research area. Vocabulary items should be of high interest to the user, be frequently applicable, have a range of meanings and be pragmatic in functionality.[5]

There are multiple methods of accessing messages on devices: directly, indirectly, and with specialized access devices, although the specific access method will depend on the skills and abilities of the user.[1] SGD output is typically much slower than speech, although rate enhancement strategies can increase the user's rate of output and as a result enhance the efficiency of communication.[6]

The first known SGD was prototyped in 1960, and rapid progress in hardware and software development has meant that SGD capabilities can now be integrated into devices like smartphones. Notable users of SGDs include Stephen Hawking, Roger Ebert, and Tony Proudfoot.

Speech generating systems may be dedicated devices developed solely for AAC, or non-dedicated devices such as computers that run additional software to allow them to function as AAC devices.[7][8]

Contents

History

The first known speech generating devices (SGD) have their roots in early electronic communication aids, the first of which was a sip-and-puff typewriter controller named the Patient Operated Selector Mechanism (POSM or POSSUM), was prototyped by Reg Maling in the United Kingdom in 1960.[9][10] POSSUM consisted of a scanning teletypewriter controller with an illuminated display.[9]. In 1970, researchers at Delft University in the Netherlands created the Lightspot Operated Typewriter (LOT), which made use of small movements of the head to point a lightspot at a matrix of characters, each equipped with a photoelectric cell. Although it was commercially unsuccessful, the LOT was well received by its users.[11]

During the 1970s and early 1980s, many of the companies that are now well known in the area began to emerge. Toby Churchill founded the company that bears his name in 1973 after losing his speech following encephalitis,[12] and Dynavox (then known as Sentient Systems Technology) grew out of a student project at Carnegie-Mellon University to help a young woman with cerebral palsy to communicate in 1982.[13] Beginning in the 1980s, improvements in technology led to a greatly increased number, range, and performance of commercially available communication devices, and a reduction in their size and price. Alternative methods of access such as eye pointing, which the movement of a user's eyes is used to direct a SGD, and scanning in which alternatives are presented to the user sequentially, became available on communication devices.[10][14] Speech output possibilities included both digitized and synthesized speech.[10]

Rapid progress in hardware and software development continued, including projects funded by the European Community. The first commercially available dynamic screen speech generating devices were developed in the 1990s and synthesized speech in more languages became available. Software programs were developed that allowed the computer-based production of communication boards.[10][14] High-tech devices have continued to become smaller and lighter,[14] while increasing accessibility and capability; communication devices can be accessed using eye-tracking systems, perform as a computer for word-processing and internet use, and as an environmental control device for independent access to other equipment such as TV, radio and telephone.[15]

As of 2011, notable AAC users include Stephen Hawking, Roger Ebert[16] and Tony Proudfoot. Hawking is unable to speak due to a combination of severe disabilities caused by ALS as well as an emergency tracheotomy.[17] He has used a DECtalk DTC01 voice synthesizer for several years,[18] and has come to be associated with the unique voice of the device. According to Hawking, he used the DTC01 for so long because he identified with it and had not heard a voice he liked better. However, he is now said to be using NeoSpeech's VoiceText speech synthesizer.[19]

Input methods

In any given SGD there may be a large number of vocal expressions that facilitate efficient and effective communication, including greetings, expressing desires, and asking questions.[20] Some SGDs have multiple pages of symbols to accommodate a large number of vocal expressions, and thus only a portion of the symbols available are visible at any one time, with the communicator navigating the various pages.[21] Speech generating devices generally display a set of selections in one of three main ways: fixed, dynamic, and hybrid.[22]

Fixed display devices

Fixed display devices refer to those in which the symbols and items are "fixed" in a particular format; some sources refer to these as "static" displays.[23] Such display devices can be simpler for users to learn to use;[23] however, this advantage quickly disappears over time as the user soon becomes familiar with it.[21]

Fixed display devices replicate the typical arrangement of low-tech AAC devices (low tech is defined as those that do not need batteries, electricity or electronics), like communication boards. They share some of their comparative disadvantages; for example they are typically restricted to a limited number of symbols, and hence messages.[22] Examples of fixed display devices include the Parakeet 15 from ZYGO industries, and DigiCom 2000 from the Great Talking Box Company.[22]

Dynamic display devices

Dynamic displays devices are usually also touchscreen devices. They typically generate electronically produced visual symbols that, when pressed, change the set of selections that is displayed. The user can change the symbols available using page links to navigate to appropriate pages of vocabulary and messages. The "home" page of a dynamic display device may show symbols related to many different contexts or conversational topics. Pressing any one of these symbols may open a different screen with messages related to that topic.[22] For example, when watching a volleyball game, a user may press the "sport" symbol to open a page with messages relating to sport, then press the symbol showing a scoreboard to utter the phrase "What's the score?".

Advantages of dynamic display devices include the availability of a much larger vocabulary, and the ability to see the sentence under construction; these enhancements must be balanced against the larger power consumption required by the display.[20] A further advantage of dynamic display devices is that the underlying operating system is capable of providing options for multiple communication channels, including cell phone, text messaging and e-mail.[24] Work by Linköping University has shown that such email writing practices allowed children who were SGD users to develop new social skills and increase their social participation.[25]

Hybrid display devices

Beukelman and Mirenda define hybrid devices as those with fixed displays that have a dynamic component – such as an LED – next to each icon that lights up when the symbol is pressed.[22] This is particularly useful in devices that employ iconic codes to represent messages, making use of a technique in which sequences of icons (pictorial symbols) are preselected and combined in order to form a word or a phrase. In such devices, different messages can be triggered by a combination of two symbols, greatly increasing the number of messages available from the display.[26] Hybrid display devices improve on fixed display devices by displaying the first symbol pressed, giving some feedback to the user before they press the second.[22]

Output

The output of a SGD may be digitized and/or synthesized: digitized systems play directly recorded words or phrases while synthesized speech uses text-to-speech software that can carry less emotional information but permits the user to speak novel messages by typing new words.[27][28] Moreover, individuals may also use a combination of recorded messages and text-to-speech techniques on their SGDs.[28] However, some devices are limited to only one type of output.

Digitized speech

Words, phrases or entire messages can be digitised and stored onto the device for playback to be activated by the user.[1][29] Such recorded speech can provide natural prosody and speech naturalness for the listener.[3] A person of the same age and gender as the AAC user can be selected to record the messages.[3]

A major disadvantage of using only recorded speech is that users are unable to produce novel messages; they are limited to the messages pre-recorded into the device.[3][30] Depending on the device, there may be a limit to the length of the recordings.[3][30]

Synthesized speech

SGDs that use synthesized speech apply the phonetic rules of the language to translate the user’s message into voice output (speech synthesis).[1][28] Users have the freedom to create novel words and messages and are not limited to those that have been pre-recorded on their device by others.[28]

Synthesized SGDs may allow multiple methods of message creation that can be used individually or in combination: messages can be created from letters, words, phrases, sentences, pictures, or symbols.[1][30] With synthesized speech there is virtually unlimited storage capacity for messages with few demands on memory space.[3]

Synthesized speech engines are available in many languages,[30][28] and the engine's parameters, such as speech rate, pitch range, gender, stress patterns, pauses, pronunciation exceptions, can be manipulated by the user.[30]

Selection set and vocabulary

Beukelman and Mirenda define a selection set of a system to be the "presentation of all messages, symbols and codes that are available at one time to a person who relies on AAC".[28] The content, organisation, and updating of this selection set is an area of active research and is influenced by a number of factors. For example, the vocabulary set for an AAC system may include words that the user does not know yet – they are included for the user to "grow into".[4] The content installed on any given SGD may include a large number of preset pages provided by the manufacturer, with a number of additional pages produced by the user or the user's care team depending on the user's needs and the contexts that the device will be used in.[4]

Initial content selection

Beukelman and Mirenda list a number of possible sources for the selection of initial content for a SGD. A range of sources is required because, in general, one individual would not have the knowledge and experience to generate all the vocal expressions needed in any given environment.[4]

Previous work has analyzed both vocabulary use of typically developing speakers and word use of AAC users to generate content for new AAC devices. Such processes work well for generating a core set of utterances or vocal expressions but are less effective in situations where a particular vocabulary is needed (for example, terms related directly to a user's interest in horse riding). The term "fringe vocabulary" refers to vocabulary that is specific or unique to the individual's personal interests or needs. A typical technique to develop fringe vocabulary for a device is to conduct interviews with multiple "informants": siblings, parents, teachers, co-workers and other involved persons.[4]

Musselwhite and St. Louis suggest that initial vocabulary items should be of high interest to the user, be frequently applicable, have a range of meanings and be pragmatic in functionality.[5] These criteria have been widely used in the AAC field as an ecological check of SGD content.[4]

Automatic content maintenance

Beukelman and Mirenda emphasize that vocabulary selection also involves ongoing vocabulary maintenance;[4] however, a bottleneck in AAC is that users or their carers must program in any new utterances manually (e.g. names of new friends or personal stories) and there are no existing commercial solutions for automatically adding content.[31] A number of research approaches have attempted to overcome this bottleneck,[32] these range from "inferred input", such as generating content based on a log of conversation with a user's friends and family,[33] to data mined from the internet to find language materials, such as the Webcrawler Project.[34] By accessing more of a user's data, more high-quality messages can be generated at a risk of exposing sensitive user data.[32] For example, by making use of global positioning systems, a device's content can be changed based on geographical location.[35][36] Moreover, by making use of Lifelogging based approaches, a device's content can be changed based on events that occur to a user during their day.[37][32]

Ethical concerns

Many recently developed SDGs include performance measurement and analysis tools to help monitor the content used by an individual. This raises concerns about privacy and the device user should be involved in the decision to monitor use in this way.[38][39] Similar concerns have been raised regarding some systems for automatic content generation,[37] and privacy is increasingly a factor in design of SGDs.[40] As AAC devices are designed to be used in all areas of a user’s life, there are sensitive legal, social, and technical issues centred on a wide family of personal data management problems that can be found in contexts of AAC use. For example, SGDs may have to be designed so that they support the user's right to delete logs of conversations or content that has been added automatically.[41]

Access methods

There are multiple methods of accessing messages on devices: directly, indirectly, and with specialized access devices. Direct access methods involve physical contact with the system, by using a keyboard or a touch screen. Users accessing SGDs indirectly and through specialized devices must manipulate an object in order to access the system, such as manoeuvring a joystick, head mouse, optical head pointer, light pointer, infrared pointer, switch access scanner, or by Morse code.[1]

The specific access method will depend on the skills and abilities of the user. With direct selection a body part, pointer, adapted mouse, joystick, or eye tracking could be used,[42] whereas switch access scanning is often used for indirect selection.[8][43] Unlike direct selection (e.g., typing on a keyboard, touching a screen), users of switch access scanning can only make selections when the scanning indicator (or cursor) of the electronic device is on the desired choice.[44] The scanning indicator moves through items by highlighting each item on the screen, or by announcing each item via voice output, and then the user activates a switch to select the item.[45] The speed and pattern of scanning, as well as the way items are selected, are individualized to the physical, visual and cognitive capabilities of the user.[44]

Rate enhancement strategies

Augmentative and alternative communication is typically much slower than speech,[6] with users generally producing 8–10 words per minute.[31] Rate enhancement strategies can increase the user's rate of output to around 12–15 words per minute,[31] and as a result enhance the efficiency of communication. There are two main options for increasing the rate of communication for SGD: encoding and prediction.[6]

Encoding permits a user to produce a word, sentence or phrase using only one or two activations of their SGD.[6] Iconic encoding strategies such as Semantic compaction combine sequences of icons (picture symbols) to produce words or phrases.[26] In numeric, alpha-numeric, and letter encoding (also known as Abbreviation-Expansion), words and sentences are coded as sequences of letters and numbers. For example, typing "HH" or "G1" (for Greeting 1) may retrieve "Hello, how are you?".[26]

Prediction is a rate enhancement strategy in which the SGD attempts to reduce the number of keystrokes used by predicting the word or phrase being written by the user. The user can then select the correct prediction without needing to write the entire word. Word prediction software may determine the choices to be offered based on their frequency in language, association with other words, past choices of the user, or grammatical suitability.[6][26][46] However, users have been shown to produce more words per minute (using a scanning interface) with a static keyboard layout than with a predictive grid layout, suggesting that the cognitive overhead of reviewing a new arrangement cancels out the benefits of the predictive layout when using a scanning interface.[47]

Another approach to rate-enhancement is Dasher,[48] which uses language models and arithmetic coding to present alternative letter targets on the screen with size relative to their likelihood given the history.[49][50]

The rate of words produced can depend greatly on the conceptual level of the system: the TALK system, which allows users to choose between large numbers of sentence-level utterances, demonstrated output rates in excess of 60 wpm.[51]

Producers

There are relatively few producers of SGDs, although several more companies produce software to give existing devices SGD functionality, and there are some home-built systems. Producers of dedicated devices include Dynavox Mayer-Johnson, Prentke Romich Company, Saltillo Corporation, and Words+.[52] Other companies produce software that allows devices like the iPhone, iPad, and Nintendo DS to function as SGDs.[53]

Notes

  1. ^ a b c d e f Aetna Inc. (2010)
  2. ^ Blischak et al (2003)
  3. ^ a b c d e f Glennen & Decoste pp. 88–90
  4. ^ a b c d e f g Beukelman & Mirenda, Chapter 2
  5. ^ a b Musselwhite & Louis
  6. ^ a b c d e University of Washington (2009)
  7. ^ Glennen, pp. 62–63.
  8. ^ a b Jans & Clark (1998), pp. 37–38.
  9. ^ a b Vanderheide (2002)
  10. ^ a b c d Zangari (1994)
  11. ^ Stassen et al., p. 127
  12. ^ Toby Churchill (About Us)
  13. ^ Dynavox (Company History
  14. ^ a b c Hourcade (2004).
  15. ^ Robitaille, pp. 151–153.
  16. ^ Chicago Sun-Times, (2009)
  17. ^ Stephen Hawking and ALS
  18. ^ Scientific American
  19. ^ Gizmag, Hawking article (2009)
  20. ^ a b Beukelman & Mirenda
  21. ^ a b Hochstein et al (2004)
  22. ^ a b c d e f Beukelman & Mirenda p. 84-85
  23. ^ a b Hochstein et al (2003)
  24. ^ Dynavox at www.speechbubble.org.uk
  25. ^ Sundqvist & Rönnberg (2010)
  26. ^ a b c d Venkatagiri (1995)
  27. ^ Schlosser, Blischak & Koul (2003)
  28. ^ a b c d e f Beukelman & Mirenda p. 105-106
  29. ^ Beukelman & Mirenda, p. 105.
  30. ^ a b c d e Radomski et al (2007)
  31. ^ a b c Higginbotham et al (2007)
  32. ^ a b c Reddington & Tintarev (2011)
  33. ^ Ashraf et al. (2002)
  34. ^ Luo et al (2007)
  35. ^ Dominowska et al
  36. ^ Patel & Radhakrishnan
  37. ^ a b Black et al (2010)
  38. ^ Beukelman & Mirenda, p. 30
  39. ^ Blackstone et al. (2002)
  40. ^ Rackensperger et al. (2005)
  41. ^ Reddington & Coles-Kemp (2011)
  42. ^ Mathy (2000)
  43. ^ Glennen & Decoste pp 62–63
  44. ^ a b Beukelman & Mirenda, pp. 97-101
  45. ^ Hedman, pp 100-101
  46. ^ Augmentative Communication, Incorporated
  47. ^ Johansen et al (2003)
  48. ^ Ward et al (2000)
  49. ^ Roark et al (2010)
  50. ^ MacKey (2003), p 119
  51. ^ Todman (2000)
  52. ^ www.infinitec.org
  53. ^ www.cbsphily.com

References